AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Subword embedding initialization

# Subword embedding initialization

Roberta Base Wechsel German
MIT
A German RoBERTa model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model Transformers German
R
benjamin
96
7
Gpt2 Wechsel French
MIT
A French version of GPT-2 trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model Transformers French
G
benjamin
33
4
Roberta Base Wechsel Swahili
MIT
A RoBERTa base model trained using the WECHSEL method, specifically optimized for Swahili to achieve efficient cross-lingual transfer.
Large Language Model Transformers Other
R
benjamin
222
1
Roberta Base Wechsel Chinese
MIT
A RoBERTa Chinese model trained with the WECHSEL method, achieving efficient cross-lingual transfer from English to Chinese
Large Language Model Transformers Chinese
R
benjamin
16
2
Gpt2 Wechsel Chinese
MIT
A Chinese GPT-2 model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model Transformers Chinese
G
benjamin
19
4
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase